Goto

Collaborating Authors

 one-shot federated learning




FedLPA: One-shot Federated Learning with Layer-Wise Posterior Aggregation

Neural Information Processing Systems

Efficiently aggregating trained neural networks from local clients into a global model on a server is a widely researched topic in federated learning. Recently, motivated by diminishing privacy concerns, mitigating potential attacks, and reducing communication overhead, one-shot federated learning (i.e., limiting client-server communication into a single round) has gained popularity among researchers. However, the one-shot aggregation performances are sensitively affected by the non-identical training data distribution, which exhibits high statistical heterogeneity in some real-world scenarios. To address this issue, we propose a novel one-shot aggregation method with layer-wise posterior aggregation, named FedLPA. FedLPA aggregates local models to obtain a more accurate global model without requiring extra auxiliary datasets or exposing any private label information, e.g., label distributions. To effectively capture the statistics maintained in the biased local datasets in the practical non-IID scenario, we efficiently infer the posteriors of each layer in each local model using layer-wise Laplace approximation and aggregate them to train the global parameters.


One-shot Federated Learning via Synthetic Distiller-Distillate Communication

Neural Information Processing Systems

One-shot Federated learning (FL) is a powerful technology facilitating collaborative training of machine learning models in a single round of communication. While its superiority lies in communication efficiency and privacy preservation compared to iterative FL, one-shot FL often compromises model performance. Prior research has primarily focused on employing data-free knowledge distillation to optimize data generators and ensemble models for better aggregating local knowledge into the server model. However, these methods typically struggle with data heterogeneity, where inconsistent local data distributions can cause teachers to provide misleading knowledge. Additionally, they may encounter scalability issues with complex datasets due to inherent two-step information loss: first, during local training (from data to model), and second, when transferring knowledge to the server model (from model to inversed data). In this paper, we propose FedSD2C, a novel and practical one-shot FL framework designed to address these challenges.


Revisiting Ensembling in One-Shot Federated Learning

Neural Information Processing Systems

Federated Learning (FL) is an appealing approach to training machine learning models without sharing raw data. However, standard FL algorithms are iterative and thus induce a significant communication cost. One-Shot FL (OFL) trades the iterative exchange of models between clients and the server with a single round of communication, thereby saving substantially on communication costs. Not surprisingly, OFL exhibits a performance gap in terms of accuracy with respect to FL, especially under high data heterogeneity. We introduce Fens, a novel federated ensembling scheme that approaches the accuracy of FL with the communication efficiency of OFL.


FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion

Neural Information Processing Systems

One-shot Federated Learning (OFL) significantly reduces communication costs in FL by aggregating trained models only once. However, the performance of advanced OFL methods is far behind the normal FL. In this work, we provide a causal view to find that this performance drop of OFL methods comes from the isolation problem, which means that local isolatedly trained models in OFL may easily fit to spurious correlations due to the data heterogeneity. From the causal perspective, we observe that the spurious fitting can be alleviated by augmenting intermediate features from other clients. Built upon our observation, we propose a novel learning approach to endow OFL with superb performance and low communication and storage costs, termed as FuseFL.


Towards One-shot Federated Learning: Advances, Challenges, and Future Directions

Amato, Flora, Qiu, Lingyu, Tanveer, Mohammad, Cuomo, Salvatore, Giampaolo, Fabio, Piccialli, Francesco

arXiv.org Artificial Intelligence

One-shot FL enables collaborative training in a single round, eliminating the need for iterative communication, making it particularly suitable for use in resource-constrained and privacy-sensitive applications. This survey offers a thorough examination of One-shot FL, highlighting its distinct operational framework compared to traditional federated approaches. One-shot FL supports resource-limited devices by enabling single-round model aggregation while maintaining data locality. The survey systematically categorizes existing methodologies, emphasizing advancements in client model initialization, aggregation techniques, and strategies for managing heterogeneous data distributions. Furthermore, we analyze the limitations of current approaches, particularly in terms of scalability and generalization in non-IID settings. By analyzing cutting-edge techniques and outlining open challenges, this survey aspires to provide a comprehensive reference for researchers and practitioners aiming to design and implement One-shot FL systems, advancing the development and adoption of One-shot FL solutions in a real-world, resource-constrained scenario.


FedBEns: One-Shot Federated Learning based on Bayesian Ensemble

Talpini, Jacopo, Savi, Marco, Neglia, Giovanni

arXiv.org Artificial Intelligence

Several One-Shot FL algorithms have been proposed in the literature. Existing relevant work leverages knowledge distillation One-Shot Federated Learning (FL) is a recent at the server (Lin et al., 2020), neuron matching paradigm that enables multiple clients to cooperatively strategies (Singh & Jaggi, 2020) or adopts an optimization learn a global model in a single round of approach, trying to directly approximate the global loss at communication with a central server. In this paper, the server starting from the local losses of each client (Jhunjhunwala we analyze the One-Shot FL problem through the et al., 2024; Liu et al., 2024; Matena & Raffel, lens of Bayesian inference and propose FedBEns, 2022). Our contribution is in line with the last group of work, an algorithm that leverages the inherent multimodality which generally employs a unimodal approximation of each of local loss functions to find better local loss. As an example, Jhunjhunwala et al. (2024) make global models.


Capture Global Feature Statistics for One-Shot Federated Learning

Guan, Zenghao, Zhou, Yucan, Gu, Xiaoyan

arXiv.org Artificial Intelligence

Traditional Federated Learning (FL) necessitates numerous rounds of communication between the server and clients, posing significant challenges including high communication costs, connection drop risks and susceptibility to privacy attacks. One-shot FL has become a compelling learning paradigm to overcome above drawbacks by enabling the training of a global server model via a single communication round. However, existing one-shot FL methods suffer from expensive computation cost on the server or clients and cannot deal with non-IID (Independent and Identically Distributed) data stably and effectively. To address these challenges, this paper proposes FedCGS, a novel Federated learning algorithm that Capture Global feature Statistics leveraging pre-trained models. With global feature statistics, we achieve training-free and heterogeneity-resistant one-shot FL. Furthermore, we extend its application to personalization scenario, where clients only need execute one extra communication round with server to download global statistics. Extensive experimental results demonstrate the effectiveness of our methods across diverse data heterogeneity settings. Code is available at https://github.com/Yuqin-G/FedCGS.


One-shot Federated Learning Methods: A Practical Guide

Liu, Xiang, Tang, Zhenheng, Li, Xia, Song, Yijun, Ji, Sijie, Liu, Zemin, Han, Bo, Jiang, Linshan, Li, Jialin

arXiv.org Artificial Intelligence

One-shot Federated Learning (OFL) is a distributed machine learning paradigm that constrains client-server communication to a single round, addressing privacy and communication overhead issues associated with multiple rounds of data exchange in traditional Federated Learning (FL). OFL demonstrates the practical potential for integration with future approaches that require collaborative training models, such as large language models (LLMs). However, current OFL methods face two major challenges: data heterogeneity and model heterogeneity, which result in subpar performance compared to conventional FL methods. Worse still, despite numerous studies addressing these limitations, a comprehensive summary is still lacking. To address these gaps, this paper presents a systematic analysis of the challenges faced by OFL and thoroughly reviews the current methods. We also offer an innovative categorization method and analyze the trade-offs of various techniques. Additionally, we discuss the most promising future directions and the technologies that should be integrated into the OFL field. This work aims to provide guidance and insights for future research.